Top-Down Cues for Event Recognition
نویسندگان
چکیده
How to fuse static and dynamic information is a key issue in event analysis. In this paper, we present a novel approach to combine appearance and motion information together through a top-down manner for event recognition in real videos. Unlike the conventional bottom-up way, attention can be focused volitionally on top-down signals derived from task demands. A video is represented by a collection of spatio-temporal features, called video words by quantizing the extracted spatio-temporal interest points (STIPs) from the video. We propose two approaches to build class specific visual or motion histograms for the corresponding features. One is using the probability of a class given a visual or motion word. High probability means more attention should be paid to this word. Moreover, in order to incorporate the negative information for each word, we propose to utilize the mutual information between each word and event label. High mutual information means high relevance between this word and the class label. Both methods not only can characterize two aspects of an event, but also can select the relevant words, which are all discriminative to the corresponding event. Experimental results on the TRECVID 2005 and the HOHA video corpus demonstrate that the mean average precision has been improved by using the proposed method.
منابع مشابه
Narrative versus Style: Effect of Genre Typical Events versus Genre Typical Filmic Realizations on Film Viewers' Genre Recognition
This study investigated whether film viewers recognize four basic genres (comic, drama, action and nonfiction) on the basis of genre-typical event cues or of genretypical filmic realization cues of events. Event cues are similar to the narrative content of a film sequence, while filmic realization cues are similar to stylistic surface cues of a film sequence. It was predicted that genre recogni...
متن کاملEvent-Related Potentials of Bottom-Up and Top-Down Processing of Emotional Faces
Introduction: Emotional stimulus is processed automatically in a bottom-up way or can be processed voluntarily in a top-down way. Imaging studies have indicated that bottom-up and top-down processing are mediated through different neural systems. However, temporal differentiation of top-down versus bottom-up processing of facial emotional expressions has remained to be clarified. The present st...
متن کاملEnhancing energy minimization framework for scene text recognition with top-down cues
Recognizing scene text is a challenging problem, even more so than the recognition of scanned documents. This problem has gained significant attention from the computer vision community in recent years, and several methods based on energy minimization frameworks and deep learning approaches have been proposed. In this work, we focus on the energy minimization framework and propose a model that ...
متن کاملObject segmentation by top-down processes
In cluttered scenes, some object boundaries may not be marked by image cues. In such cases, the boundaries must be defined top-down as a result of object recognition. Here we ask if observers can retain the boundaries of several recognized objects in order to segment an unfamiliar object. We generated scenes consisting of neatly stacked objects, and the objects themselves consisted of neatly st...
متن کاملFirst steps towards autonomous recognition of Monterey Bay’s most common mid-water organisms: Mining the ROV video database on behalf of the Automated Visual Event Detection (AVED) system
The development of remotely operated vehicles (ROVs) has revolutionized the world of marine science by allowing quantitative video transects (QVTs) to be recorded underwater. This non-invasive technique allows previously unavailable information to be obtained concerning organism diversity, distribution, and abundance. Unfortunately, processing of these QVTs at the Monterey Bay Aquarium Research...
متن کاملذخیره در منابع من
با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید
عنوان ژورنال:
دوره شماره
صفحات -
تاریخ انتشار 2010